经典的媒体访问控制(MAC)协议是可解释的,但是它们的任务不可能控制信号传导消息(CMS)不适合新兴任务 - 关键任务应用程序。相比之下,基于神经网络(NN)协议模型(NPM)学会生成特定于任务的CMS,但其理由和影响缺乏可解释性。为了填补这一空白,在本文中,我们首次提出了通过将NPM转换为概率逻辑编程语言(ProBlog)编写的可解释的符号图来构建的语义协议模型(SPM)。通过在将NPM视为CM发生器的同时提取和合并共同的CM及其连接,可以可行。通过广泛的模拟,我们证实了SPM在仅占据0.02%内存的同时紧密近似其原始NPM。通过利用其可解释性和记忆效率,我们演示了几种支持SPM的应用程序,例如SPM重新配置,以避免碰撞,并通过语义熵计算和存储多个SPM来比较不同的SPM,以应对非平稳环境。
translated by 谷歌翻译
许多机器人任务由在高度复杂的环境中由许多时间相关的子任务组成。重要的是要通过审议时间抽象来有效地解决问题来发现情境意图和适当的行动。为了了解与不断变化的任务动态分离的意图,我们将基于授权的正则化技术扩展到基于生成对抗网络框架的多个任务的情况。在具有未知动态的多任务环境下,我们专注于从未标记的专家示例中学习奖励和政策。在这项研究中,我们将情境增强权定义为相互信息的最大信息,代表了在某个状态和子任务中如何影响未来的行动。我们提出的方法得出了情境相互信息的变异下限,以优化它。我们同时通过在目标函数中添加引起的术语来同时学习可转让的多任务奖励功能和策略。通过这样做,多任务奖励功能有助于学习对环境变化的强大政策。我们验证了我们在多任务学习和多任务转移学习方面的优势。我们证明我们提出的方法具有随机性和变化的任务动态的鲁棒性。最后,我们证明我们的方法的性能和数据效率明显优于各种基准上的现有模仿学习方法。
translated by 谷歌翻译
应用强化学习来自动驾驶需要某些挑战,这主要是由于大规模的交通流动,这种挑战是动态变化的。为了应对此类挑战,有必要快速确定对周围车辆不断变化的意图的响应策略。因此,我们提出了一种新的政策优化方法,用于使用基于图的互动感知约束来安全驾驶。在此框架中,运动预测和控制模块是同时训练的,同时共享包含社会环境的潜在表示。此外,为了反映社交互动,我们以图形形式表达了代理的运动并过滤特征。这有助于保留相邻节点的时空位置。此外,我们创建反馈循环以有效地组合这两个模块。结果,这种方法鼓励博学的控制器免受动态风险的侵害,并在各种情况下使运动预测强大。在实验中,我们与城市驾驶模拟器Carla建立了一个包括各种情况的导航场景。该实验表明,与基线相比,导航策略和运动预测的两侧的最新性能。
translated by 谷歌翻译
In this paper, we propose a diffusion-based face swapping framework for the first time, called DiffFace, composed of training ID conditional DDPM, sampling with facial guidance, and a target-preserving blending. In specific, in the training process, the ID conditional DDPM is trained to generate face images with the desired identity. In the sampling process, we use the off-the-shelf facial expert models to make the model transfer source identity while preserving target attributes faithfully. During this process, to preserve the background of the target image and obtain the desired face swapping result, we additionally propose a target-preserving blending strategy. It helps our model to keep the attributes of the target face from noise while transferring the source facial identity. In addition, without any re-training, our model can flexibly apply additional facial guidance and adaptively control the ID-attributes trade-off to achieve the desired results. To the best of our knowledge, this is the first approach that applies the diffusion model in face swapping task. Compared with previous GAN-based approaches, by taking advantage of the diffusion model for the face swapping task, DiffFace achieves better benefits such as training stability, high fidelity, diversity of the samples, and controllability. Extensive experiments show that our DiffFace is comparable or superior to the state-of-the-art methods on several standard face swapping benchmarks.
translated by 谷歌翻译
For change detection in remote sensing, constructing a training dataset for deep learning models is difficult due to the requirements of bi-temporal supervision. To overcome this issue, single-temporal supervision which treats change labels as the difference of two semantic masks has been proposed. This novel method trains a change detector using two spatially unrelated images with corresponding semantic labels such as building. However, training on unpaired datasets could confuse the change detector in the case of pixels that are labeled unchanged but are visually significantly different. In order to maintain the visual similarity in unchanged area, in this paper, we emphasize that the change originates from the source image and show that manipulating the source image as an after-image is crucial to the performance of change detection. Extensive experiments demonstrate the importance of maintaining visual information between pre- and post-event images, and our method outperforms existing methods based on single-temporal supervision. code is available at https://github.com/seominseok0429/Self-Pair-for-Change-Detection.
translated by 谷歌翻译
In recent years, generative models have undergone significant advancement due to the success of diffusion models. The success of these models is often attributed to their use of guidance techniques, such as classifier and classifier-free methods, which provides effective mechanisms to trade-off between fidelity and diversity. However, these methods are not capable of guiding a generated image to be aware of its geometric configuration, e.g., depth, which hinders the application of diffusion models to areas that require a certain level of depth awareness. To address this limitation, we propose a novel guidance approach for diffusion models that uses estimated depth information derived from the rich intermediate representations of diffusion models. To do this, we first present a label-efficient depth estimation framework using the internal representations of diffusion models. At the sampling phase, we utilize two guidance techniques to self-condition the generated image using the estimated depth map, the first of which uses pseudo-labeling, and the subsequent one uses a depth-domain diffusion prior. Experiments and extensive ablation studies demonstrate the effectiveness of our method in guiding the diffusion models toward geometrically plausible image generation. Project page is available at https://ku-cvlab.github.io/DAG/.
translated by 谷歌翻译
Deep learning-based weather prediction models have advanced significantly in recent years. However, data-driven models based on deep learning are difficult to apply to real-world applications because they are vulnerable to spatial-temporal shifts. A weather prediction task is especially susceptible to spatial-temporal shifts when the model is overfitted to locality and seasonality. In this paper, we propose a training strategy to make the weather prediction model robust to spatial-temporal shifts. We first analyze the effect of hyperparameters and augmentations of the existing training strategy on the spatial-temporal shift robustness of the model. Next, we propose an optimal combination of hyperparameters and augmentation based on the analysis results and a test-time augmentation. We performed all experiments on the W4C22 Transfer dataset and achieved the 1st performance.
translated by 谷歌翻译
Traditional weather forecasting relies on domain expertise and computationally intensive numerical simulation systems. Recently, with the development of a data-driven approach, weather forecasting based on deep learning has been receiving attention. Deep learning-based weather forecasting has made stunning progress, from various backbone studies using CNN, RNN, and Transformer to training strategies using weather observations datasets with auxiliary inputs. All of this progress has contributed to the field of weather forecasting; however, many elements and complex structures of deep learning models prevent us from reaching physical interpretations. This paper proposes a SImple baseline with a spatiotemporal context Aggregation Network (SIANet) that achieved state-of-the-art in 4 parts of 5 benchmarks of W4C22. This simple but efficient structure uses only satellite images and CNNs in an end-to-end fashion without using a multi-model ensemble or fine-tuning. This simplicity of SIANet can be used as a solid baseline that can be easily applied in weather forecasting using deep learning.
translated by 谷歌翻译
Traversability estimation for mobile robots in off-road environments requires more than conventional semantic segmentation used in constrained environments like on-road conditions. Recently, approaches to learning a traversability estimation from past driving experiences in a self-supervised manner are arising as they can significantly reduce human labeling costs and labeling errors. However, the self-supervised data only provide supervision for the actually traversed regions, inducing epistemic uncertainty according to the scarcity of negative information. Negative data are rarely harvested as the system can be severely damaged while logging the data. To mitigate the uncertainty, we introduce a deep metric learning-based method to incorporate unlabeled data with a few positive and negative prototypes in order to leverage the uncertainty, which jointly learns using semantic segmentation and traversability regression. To firmly evaluate the proposed framework, we introduce a new evaluation metric that comprehensively evaluates the segmentation and regression. Additionally, we construct a driving dataset `Dtrail' in off-road environments with a mobile robot platform, which is composed of a wide variety of negative data. We examine our method on Dtrail as well as the publicly available SemanticKITTI dataset.
translated by 谷歌翻译
This work presents six structural quality metrics that can measure the quality of knowledge graphs and analyzes five cross-domain knowledge graphs on the web (Wikidata, DBpedia, YAGO, Google Knowledge Graph, Freebase) as well as 'Raftel', Naver's integrated knowledge graph. The 'Good Knowledge Graph' should define detailed classes and properties in its ontology so that knowledge in the real world can be expressed abundantly. Also, instances and RDF triples should use the classes and properties actively. Therefore, we tried to examine the internal quality of knowledge graphs numerically by focusing on the structure of the ontology, which is the schema of knowledge graphs, and the degree of use thereof. As a result of the analysis, it was possible to find the characteristics of a knowledge graph that could not be known only by scale-related indicators such as the number of classes and properties.
translated by 谷歌翻译